📚 node [[return|return]]
Welcome! Nobody has contributed anything to 'return|return' yet. You can:
-
Write something in the document below!
- There is at least one public document in every node in the Agora. Whatever you write in it will be integrated and made available for the next visitor to read and edit.
- Write to the Agora from social media.
-
Sign up as a full Agora user.
- As a full user you will be able to contribute your personal notes and resources directly to this knowledge commons. Some setup required :)
⥅ related node [[2003 12 17 return of the awesome]]
⥅ related node [[return]]
⥅ related node [[online platforms should return value to the citizen body]]
⥅ related node [[20200607142108 roberts_returning_to_normal]]
⥅ node [[return]] pulled by Agora
📓
garden/KGBicheno/Artificial Intelligence/Introduction to AI/Week 3 - Introduction/Definitions/Return.md by @KGBicheno
return
Go back to the [[AI Glossary]]
#rl
In reinforcement learning, given a certain policy and a certain state, the return is the sum of all rewards that the agent expects to receive when following the policy from the state to the end of the episode. The agent accounts for the delayed nature of expected rewards by discounting rewards according to the state transitions required to obtain the reward.
$$\text {Therefore, if the discount factor is } \lambda \text{, and } r_o,\ldots,r_n \text{ denote the rewards until the end of the episode, then the return calculation is as follows:} $$
📖 stoas
- public document at doc.anagora.org/return|return
- video call at meet.jit.si/return|return
🔎 full text search for 'return|return'